AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Human preference optimization

# Human preference optimization

Btlm 3b 8k Chat
Apache-2.0
BTLM-3B-8k-chat is a conversational version developed based on BTLM-3B-8K-base, optimized using DPO method, specifically designed for dialogue scenarios aligned with human preferences.
Large Language Model Transformers English
B
cerebras
138
13
7B DPO Alpha
A 7B-parameter causal language model trained on multi-source datasets, optimized with DPO, supporting Chinese and English text generation tasks
Large Language Model Transformers Supports Multiple Languages
7
CausalLM
131
54
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase